Statistical consistency and asymptotic normality for high-dimensional robust M-estimators
نویسنده
چکیده
We study theoretical properties of regularized robust M -estimators, applicable when data are drawn from a sparse high-dimensional linear model and contaminated by heavytailed distributions and/or outliers in the additive errors and covariates. We first establish a form of local statistical consistency for the penalized regression estimators under fairly mild conditions on the error distribution: When the derivative of the loss function is bounded and satisfies a local restricted curvature condition, all stationary points within a constant radius of the true regression vector converge at the minimax rate enjoyed by the Lasso with sub-Gaussian errors. When an appropriate nonconvex regularizer is used in place of an l1-penalty, we show that such stationary points are in fact unique and equal to the local oracle solution with the correct support—hence, results on asymptotic normality in the low-dimensional case carry over immediately to the high-dimensional setting. This has important implications for the efficiency of regularized nonconvex M estimators when the errors are heavy-tailed. Our analysis of the local curvature of the loss function also has useful consequences for optimization when the robust regression function and/or regularizer is nonconvex and the objective function possesses stationary points outside the local region. We show that as long as a composite gradient descent algorithm is initialized within a constant radius of the true regression vector, successive iterates will converge at a linear rate to a stationary point within the local region. Furthermore, the global optimum of a convex regularized robust regression function may be used to obtain a suitable initialization. The result is a novel two-step procedure that uses a convex M estimator to achieve consistency and a nonconvex M -estimator to increase efficiency. We conclude with simulation results that corroborate our theoretical findings.
منابع مشابه
Nonconcave Penalized M-estimation with a Diverging Number of Parameters
M-estimation is a widely used technique for robust statistical inference. In this paper, we investigate the asymptotic properties of a nonconcave penalized M-estimator in sparse, high-dimensional, linear regression models. Compared with classic M-estimation, the nonconcave penalized M-estimation method can perform parameter estimation and variable selection simultaneously. The proposed method i...
متن کاملAsymptotic Behaviors of Nearest Neighbor Kernel Density Estimator in Left-truncated Data
Kernel density estimators are the basic tools for density estimation in non-parametric statistics. The k-nearest neighbor kernel estimators represent a special form of kernel density estimators, in which the bandwidth is varied depending on the location of the sample points. In this paper, we initially introduce the k-nearest neighbor kernel density estimator in the random left-truncatio...
متن کاملOn discrete-domain multidimensional sinusoidal models
We consider three-dimensional sinusoidal frequency model in a random field. Three-dimensional frequency model has wide applications in statistical signal processing. In this article, we mainly consider the usual least squares estimators and the estimators that can be obtained by maximizing the periodogram function. We obtain consistency and asymptotic normality property of both the estimators. ...
متن کاملRobust semiparametric M-estimation and the weighted bootstrap
M-estimation is a widely used technique for statistical inference. In this paper, we study properties of ordinary and weighted M-estimators for semiparametric models, especially when there exist parameters that cannot be estimated at the √ n convergence rate. Results on consistency, rates of convergence for all parameters, and √ n consistency and asymptotic normality for the Euclidean parameter...
متن کاملON HADAMARD DIFFERENTIABILITY AND M-ESTIMATION IN LINEAR MODELS by
Robust (M-) estimation in linear models generally involves statistical functional processes. For drawing statistical conclusions (in large samples), some (uniform) linear approximations are usually needed for such functionals. In this context, the role of Hadamard differentiability is critically examined in this dissertation. In particular, the concept of the second-order Hadamard differenti-ab...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
- CoRR
دوره abs/1501.00312 شماره
صفحات -
تاریخ انتشار 2015